equalized odds
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- North America > Canada (0.04)
- Banking & Finance (0.68)
- Government (0.46)
- North America > United States > Illinois > Cook County > Chicago (0.04)
- North America > United States > Pennsylvania > Philadelphia County > Philadelphia (0.04)
- North America > Greenland (0.04)
- (10 more...)
- Law (1.00)
- Government (0.93)
- North America > United States > Illinois > Cook County > Chicago (0.05)
- North America > United States > California (0.04)
- North America > Canada > Quebec > Montreal (0.04)
d3222559698f41247261b7a6c2bbaedc-Paper-Conference.pdf
The impossibility theorem of fairness is a foundational result in the algorithmic fairness literature. It states that outside of special cases, one cannot exactly and simultaneously satisfy all three common and intuitive definitions of fairness demographic parity, equalized odds, and predictive rate parity. This result has driven most works to focus on solutions for one or two of the metrics.
- North America > United States > California > Santa Clara County > Sunnyvale (0.05)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- Europe > Germany (0.04)
ad991bbc381626a8e44dc5414aa136a8-Supplemental-Conference.pdf
Figure 1 shows the change of accuracy under different cutoff value. However, for gender classification under CelebA dataset, thetrade-offbetweenλval and accuracyisnotveryclear;and wesuspect that under suchscenario, focusing on hard samples does not harm the performance of easy samples, and thus benefits the classifier. Figure 1 shows the change of fairness (equalized odds) under different cutoff value. Suppose we have a large unlabeled training set of sizeN and a small labeled validation set { xvalj,yvalj,1 j M} with M N. In each training step, we sample a small mini-batch of size n(n < N) from training set and perform random augmentation twice to obtain a subset { xi,1 i 2n} and we update the contrastive encoderf with parameterθ. During validation, we freeze the contrastive encoder and train a downstream linear classifierg with parameterω for classification task.
- North America > United States > Virginia (0.04)
- North America > United States > New Mexico > Bernalillo County > Albuquerque (0.04)
- North America > United States > Georgia > Fulton County > Atlanta (0.04)
- (4 more...)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
- North America > United States > California (0.04)
- North America > Canada (0.04)
- North America > United States > California > Santa Clara County > Stanford (0.04)
- North America > United States > California > Santa Clara County > Palo Alto (0.04)
- North America > United States > Arizona (0.04)
- North America > Canada (0.04)
- Information Technology > Data Science (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (0.94)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.68)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.47)
Estimating and Controlling for Equalized Odds via Sensitive Attribute Predictors
As the use of machine learning models in real world high-stakes decision settings continues to grow, it is highly important that we are able to audit and control for any potential fairness violations these models may exhibit towards certain groups. To do so, one naturally requires access to sensitive attributes, such as demographics, biological sex, or other potentially sensitive features that determine group membership. Unfortunately, in many settings, this information is often unavailable. In this work we study the well known equalized odds (EOD) definition of fairness. In a setting without sensitive attributes, we first provide tight and computable upper bounds for the EOD violation of a predictor. These bounds precisely reflect the worst possible EOD violation. Second, we demonstrate how one can provably control the worst-case EOD by a new post-processing correction method. Our results characterize when directly controlling for EOD with respect to the predicted sensitive attributes is -- and when is not -- optimal when it comes to controlling worst-case EOD. Our results hold under assumptions that are milder than previous works, and we illustrate these results with experiments on synthetic and real datasets.